161 research outputs found

    Gabor-enhanced histogram of oriented gradients for human presence detection applied in aerial monitoring

    Get PDF
    In UAV-based human detection, the extraction and selection of the feature vector are one of the critical tasks to ensure the optimal performance of the detection system. Although UAV cameras capture high-resolution images, human figures' relative size renders persons at very low resolution and contrast. Feature descriptors that can adequately discriminate between local symmetrical patterns in a low-contrast image may improve a human figures' detection in vegetative environments. Such a descriptor is proposed and presented in this paper. Initially, the acquired images are fed to a digital processor in a ground station where the human detection algorithm is performed. Part of the human detection algorithm is the GeHOG feature extraction, where a bank of Gabor filters is used to generate textured images from the original. The local energy for each cell of the Gabor images is calculated to identify the dominant orientations. The bins of conventional HOG are enhanced based on the dominant orientation index and the accumulated local energy in Gabor images. To measure the performance of the proposed features, Gabor-enhanced HOG (GeHOG) and other two recent improvements to HOG, Histogram of Edge Oriented Gradients (HEOG) and Improved HOG (ImHOG), are used for human detection on INRIA dataset and a custom dataset of farmers working in fields captured via unmanned aerial vehicle. The proposed feature descriptor significantly improved human detection and performed better than recent improvements in conventional HOG. Using GeHOG improved the precision of human detection to 98.23% in the INRIA dataset. The proposed feature can significantly improve human detection applied in surveillance systems, especially in vegetative environments

    Alignment control using visual servoing and mobilenet single-shot multi-box detection (SSD): a review

    Get PDF
    The concept is highly critical for robotic technologies that rely on visual feedback. In this context, robot systems tend to be unresponsive due to reliance on pre-programmed trajectory and path, meaning the occurrence of a change in the environment or the absence of an object. This review paper aims to provide comprehensive studies on the recent application of visual servoing and DNN. PBVS and Mobilenet-SSD were chosen algorithms for alignment control of the film handler mechanism of the portable x-ray system. It also discussed the theoretical framework features extraction and description, visual servoing, and Mobilenet-SSD. Likewise, the latest applications of visual servoing and DNN was summarized, including the comparison of Mobilenet-SSD with other sophisticated models. As a result of a previous study presented, visual servoing and MobileNet-SSD provide reliable tools and models for manipulating robotics systems, including where occlusion is present. Furthermore, effective alignment control relies significantly on visual servoing and deep neural reliability, shaped by different parameters such as the type of visual servoing, feature extraction and description, and DNNs used to construct a robust state estimator. Therefore, visual servoing and MobileNet-SSD are parameterized concepts that require enhanced optimization to achieve a specific purpose with distinct tools

    Object Detection in X-ray Images Using Transfer Learning with Data Augmentation

    Get PDF
    Object detection in X-ray images is an interesting problem in the field of machine vision. The reason is that images from an X-ray machine are usually obstructed with other objects and to itself, therefore object classification and localization is a challenging task. Furthermore, obtaining X-ray data is difficult due to an insufficient dataset available compared with photographic images from a digital camera. It is vital to easily detect objects in an X-ray image because it can be used as decision support in the detection of threat items such as improvised explosive devices (IED’s) in airports, train stations, and public places. Detection of IED components accurately requires an expert and can be achieved through extensive training. Also, manual inspection is tedious, and the probability of missed detection increases due to several pieces of baggage are scanned in a short period of time. As a solution, this paper used different object detection techniques (Faster R-CNN, SSD, R-FCN) and feature extractors (ResNet, MobileNet, Inception, Inception-ResNet) based on convolutional neural networks (CNN) in a novel IEDXray dataset in the detection of IED components. The IEDXray dataset is an X-ray image of IED replicas without the explosive material. Transfer learning with data augmentation was performed due to limited X-ray data available to train the whole network from scratch. Evaluation results showed that individual detection achieved 99.08% average precision (AP) in mortar detection and 77.29% mAP in three IED components

    Optimization of CO2 Laser Cutting Parameters Using Adaptive Neuro-Fuzzy Inference System (ANFIS)

    Get PDF
    Laser cutting is a manufacturing technology that uses laser light to cut almost any materials. This type of cutting technology has been applied in many industrial applications. Problems seen with a laser is the cutting efficiency and the quality wherein these two parameters are both affected by the laser power and its process speed. This study presents the modelling and simulation of an intelligent system for predicting and optimising the process parameters of CO2 laser cutting. The developed model was trained and tested using actual data gathered from actual laser cut runs. For the system parameters, two inputs were used: the type of material used and the material thickness (mm). For the desired response, the output is the process speed or cutting rate (mm/min). Adaptive neuro-fuzzy inference system (ANFIS) was the tool used to model the optimisation cutting process. Moreover, grid partition (GP) and subtractive clustering were both used in designing the fuzzy inference system (FIS). Among the training models used, GP Gaussian bell membership function (Gbellmf) provided the highest performance with an accuracy of 99.66%

    Lettuce growth stage identification based on phytomorphological variations using coupled color superpixels and multifold watershed transformation

    Get PDF
    Identifying the plant's developmental growth stages from seed leaf is crucial to understand plant science and cultivation management deeply. An efficient vision-based system for plant growth monitoring entails optimum segmentation and classification algorithms. This study presents coupled color-based superpixels and multifold watershed transformation in segmenting lettuce plant from complicated background taken from smart farm aquaponic system, and machine learning models used to classify lettuce plant growth as vegetative, head development and for harvest based on phytomorphological profile. Morphological computations were employed by feature extraction of the number of leaves, biomass area and perimeter, convex area, convex hull area and perimeter, major and minor axis lengths of the major axis length the dominant leaf, and length of plant skeleton. Phytomorphological variations of biomass compactness, convexity, solidity, plant skeleton, and perimeter ratio were included as inputs of the classification network. The extracted Lab color space information from the training image set undergoes superpixels overlaying with 1,000 superpixel regions employing K-means clustering on each pixel class. Six-level watershed transformation with distance transformation and minima imposition was employed to segment the lettuce plant from other pixel objects. The accuracy of correctly classifying the vegetative, head development, and harvest growth stages are 88.89%, 86.67%, and 79.63%, respectively. The experiment shows that the test accuracy rates of machine learning models were recorded as 60% for LDA, 85% for ANN, and 88.33% for QSVM. Comparative analysis showed that QSVM bested the performance of optimized LDA and ANN in classifying lettuce growth stages. This research developed a seamless model in segmenting vegetation pixels, and predicting lettuce growth stage is essential for plant computational phenotyping and agricultural practice optimization

    Optimization of Aquaponic Lettuce Evapotranspiration Based on Artificial Photosynthetic Light Properties Using Hybrid Genetic Programming and Moth Flame Optimizer

    Get PDF
    Land and water resources, climate change, and disaster risks significantly affect the agricultural sector. An effective solution for growing crops to improve productivity and optimize the use of resources is through controlled-environment agriculture (CEA). Evapotranspiration (ET) is an important greenhouse crop attribute that can be optimized for optimum plant growth. Light intensity and radiation are significant for controlling ET. To address this challenge, this study successfully determined the properties of optimum artificial light for minimum evapotranspiration rate of head development-stage and harvest-stage lettuce under light-period and dark-period using genetic programming and bio-inspired algorithms namely, grey wolf optimization (GWO), whale optimization algorithm (WOA), dragonfly algorithm (DA), and moth flame optimization (MFO). MFO provided the optimized global solution for the configured models. Results showed that head development-stage lettuce requires higher light intensity with lower visible to infrared radiation ratio (Vis/IR) than harvest-stage lettuce when exposed to light. On the other hand, harvest-stage lettuce requires higher light intensity with lower Vis/IR than head development-stage under dark-period respiration reaction. Findings of this study can be utilized in growing and improving yield crops in controlled-environment agriculture

    Automation and Control for Adaptive Management System of Urban Agriculture Using Computational Intelligence

    Get PDF
    It has been predicted by the United Nations that the world population will increase to 9.8 billion in 2050. This causes agricultural development areas to be transformed into urban areas. This urbanization and increase in population density cause food insecurity. Urban agriculture using precision farming becomes a feasible solution to meet the growing demand for food and space. An adaptive management system (AMS) is necessary for such farm to provide an artificial environment suitable to produce cultivars effectively. This research proposes the development of a computational intelligence-based urban farm automation and control system utilizing machine learning and fuzzy logic system models. A quality assessment is employed for adjusting the environmental parameters with respect to the cultivars’ requirements. The system is composed of sensors for data acquisition and actuators for model-dictated responses to stimuli. Data logging was done wirelessly through a router that would collect and monitor data through a cloud-based dashboard. The model intended for training from the acquired data undergo statistical comparative analysis and least computational cost analysis to optimize the performance. The system performance was evaluated by monitoring the conditions of the sensors and actuators. Experiment results showed that the proposed system is accurate, robust, and reliable

    Adaptive Neuro-Fuzzy Inference System-Based GPS-IMU Data Correction for Capacitive Resistivity Underground Imaging with Towed Vehicle System

    Get PDF
    This study proposes the utilization of an Adaptive Neuro-Fuzzy Inference System (ANFIS) to correct the latitude and longitude of Global Positioning System (GPS) used in locating towed vehicle system for underground imaging. The input used was the collected data from a developed Real-time Kinematic Global Positioning System sensor integrated with Inertial Measurement Unit. Different ANFIS models were developed and evaluated. For latitude correction, ANFIS model with hybrid optimization trained at 300 epochs was chosen, whereas for longitude correction, ANFIS model with hybrid optimization trained at 100 epochs was selected. Both models achieved the lowest Mean Squared Error (MSE), the highest Coefficient of Determination (R2), and lowest Mean Absolute Error (MAE). Moreover, selected best ANFIS models were compared to Long Short-Term Memory (LSTM) and Extreme Learning Machine (ELM) models, but the results showed that the ANFIS models have superior performances. The selected ANFIS models were verified by testing on the collected actual dataset and the visualized map demonstrated that the corrected GPS latitude and longitude have significantly reduced error, indicating that the fuzzy system with neural network capabilities is a cost-effective and convenient method for error reduction in vehicle localization making it applicable to be integrated for capacitive resistivity underground imaging systems

    Stereo Vision 3D Tracking of Multiple Free-Swimming Fish for Low Frame Rate Video

    Get PDF
    3D multiple fish tracking has gained a significant growing research interest to quantify fish behavior. However, most tracking techniques have used a high frame rate that is currently not viable for real-time tracking applications. This study discusses multiple fish tracking techniques using low frame rate sampling of stereo video clips. The fish are tagged and tracked based on the absolute error of predicted indices using past and present fish centroid locations and a deterministic frame index. In the predictor sub-system, the linear regression and machine learning algorithms intended for nonlinear systems, such as Adaptive Neuro-Fuzzy Inference System (ANFIS), symbolic regression, and Gaussian Process Regression (GPR), were investigated. Results have shown that in the context of tagging and tracking accuracy, the symbolic regression attained the best performance, followed by the GPR, i.e., 74% to 100% and 81% to 91%, respectively. Considering the computation time, symbolic regression resulted in the highest computing lag of approximately 946 ms per iteration, whereas GPR achieved the lowest computing time of 39 ms

    Development and design of mobile robot with IP-based vision system

    No full text
    A hardware, firmware and software design of a mobile robot capable of transmitting video information and receiving commands from a controlling point is presented. The hardware design is composed of a PIC18F4620 microcontroller, a UCC27525 MOSFET gate driver, XBee Series 2 OEM RF Module. Firmware design includes the reception, processing and decoding of Zigbee API packets. Based on this decoded information the microcontroller will generate signals to move the motors namely left and right motors with a corresponding direction, either clockwise of counterclockwise. The software part includes the graphical user interface which generates commands sent to the mobile robot. The images from the mobile robot is sent to the central controller. The images are then processed and a command is generated. The command is formatted in API format and then sent to the mobile robot. Testing of the system is done by experimentation. Three parameters are tested which are influenced by four parameters. Image recognition is measured while varying the distance. Also image recognition is measured while varying the luminance of the environment. The received signal level is measured while varying the distance. Lastly the accuracy of the movement of the mobile robot is also measured while varying the target position. The results showed that the distance used by the researcher has no significant effect on image recognition. The results showed also that image recognition is unaffected vi Development and Design of Mobile Robot with IP-based Vision System with the luminance of 230-1590 lumens. The mobile robot will respond in an optimum range of one meter and can respond from one to ten meters
    • …
    corecore